Contextual Inverse Optimization: Offline and Online Learning

نویسندگان

چکیده

We study the problems of offline and online contextual optimization with feedback information, where instead observing loss, we observe, after-the-fact, optimal action an oracle full knowledge objective function would have taken. aim to minimize regret, which is defined as difference between our losses ones incurred by all-knowing oracle. In setting, decision-maker has information available from past periods needs make one decision, while in optimizes decisions dynamically over time based a new set feasible actions functions each period. For characterize minimax policy, establishing performance that can be achieved underlying geometry induced data. leverage this geometric characterization optimize cumulative regret. develop algorithm yields first regret bound for problem logarithmic horizon.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Offline Optimization for Online Ad Allocation

We consider online ad allocation from the perspective of optimizing delivery of a given set of ad reservations. Ad allocation naturally fits into the class of online bipartite matching problems: ad nodes are fixed, and impression (ad slot) nodes arrive one at a time, and must be assigned a qualified ad upon arrival. However most previous work on online matching does not model an important aspec...

متن کامل

Online Optimization through Mining the Offline Optimum

Ports, warehouses and courier services have to decide online how an arriving task is to be served in order that cost is minimized (or profit maximized). These operators have a wealth of historical data on task assignments; can these data be mined for knowledge or rules that can help the decisionmaking? MOO is a novel application of data mining to online optimization. The idea is to mine (logged...

متن کامل

Emulating the Expert: Inverse Optimization through Online Learning

In this paper, we demonstrate how to learn the objective function of a decision maker while only observing the problem input data and the decision maker’s corresponding decisions over multiple rounds. Our approach is based on online learning techniques and works for linear objectives over arbitrary sets for which we have a linear optimization oracle and as such generalizes previous work based o...

متن کامل

Online Heterogeneous Transfer Learning by Weighted Offline and Online Classifiers

In this paper, we study online heterogeneous transfer learning (HTL) problems where offline labeled data from a source domain is transferred to enhance the online classification performance in a target domain. The main idea of our proposed algorithm is to build an offline classifier based on heterogeneous similarity constructed by using labeled data from a source domain and unlabeled co-occurre...

متن کامل

Offline Evaluation of Online Reinforcement Learning Algorithms

In many real-world reinforcement learning problems, we have access to an existing dataset and would like to use it to evaluate various learning approaches. Typically, one would prefer not to deploy a fixed policy, but rather an algorithm that learns to improve its behavior as it gains more experience. Therefore, we seek to evaluate how a proposed algorithm learns in our environment, meaning we ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Social Science Research Network

سال: 2021

ISSN: ['1556-5068']

DOI: https://doi.org/10.2139/ssrn.3863366